Going Wider: Recurrent Neural Network With Parallel Cells

نویسندگان

  • Danhao Zhu
  • Si Shen
  • Xin-Yu Dai
  • Jiajun Chen
چکیده

Recurrent Neural Network (RNN) has been widely applied for sequence modeling. In RNN, the hidden states at current step are full connected to those at previous step, thus the influence from less related features at previous step may potentially decrease model’s learning ability. We propose a simple technique called parallel cells (PCs) to enhance the learning ability of Recurrent Neural Network (RNN). In each layer, we run multiple small RNN cells rather than one single large cell. In this paper, we evaluate PCs on 2 tasks. On language modeling task on PTB (Penn Tree Bank), our model outperforms state of art models by decreasing perplexity from 78.6 to 75.3. On Chinese-English translation task, our model increases BLEU score for 0.39 points than baseline model. 1.Introduction Recurrent neural network (RNN) has been one of the most powerful sequence models in natural language processing. Many important applications achieve state of the art performance with RNN, including language modeling(Mikolov et al.,2010; Zaremba et al., 2014 ), machine translation (Luong and Manning, 2016; Wu et al., 2016) and so on. The features learned by RNN are stored in the hidden states. At each time step, the cell extracts features from data and updates its hidden states. The left side of figure 1 concisely shows the transition of hidden states in naïve RNN. We can see that all units at the previous step are fully connected to all units at the current step. Thus, each pair of features can affect each other. Such design is not reality because many features are not that related. The influence between unrelated features may harm the learning ability of models. We can expect learning models automatically set the weight of all unnecessary connections to zero. However, in practice, because the data size is limited and the algorithms are not that strong, these unrelated connections will harm the learning ability. For example, that is why we have to do feature selections before training. To address the problem, many successful neural models benefit from replacing global connection with local connection. For example, Long Short-Term Memory(LSTM) (Hochreiter and Schmidhuber, 1997) and Gated Recurrent Units(GRU) (Cho et al., 2014) have been the most popular RNN cells. In the core, such models use gates to control the data flow, allow part of connections to be activated. Another example is Convolution Neural Networks(CNN)(Lecun et al, 1998), one of the most successful models in deep learning nowadays. CNN uses local receptive fields to extract features from previous feature map. With local receptive fields, neurons can extract elementary visual features such as oriented edges, end-points, corners.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A New Recurrent Fuzzy Neural Network Controller Design for Speed and Exhaust Temperature of a Gas Turbine Power Plant

In this paper, a recurrent fuzzy-neural network (RFNN) controller with neural network identifier in direct control model is designed to control the speed and exhaust temperature of the gas turbine in a combined cycle power plant. Since the turbine operation in combined cycle unit is considered, speed and exhaust temperature of the gas turbine should be simultaneously controlled by fuel command ...

متن کامل

Forward kinematic analysis of planar parallel robots using a neural network-based approach optimized by machine learning

The forward kinematic problem of parallel robots is always considered as a challenge in the field of parallel robots due to the obtained nonlinear system of equations. In this paper, the forward kinematic problem of planar parallel robots in their workspace is investigated using a neural network based approach. In order to increase the accuracy of this method, the workspace of the parallel robo...

متن کامل

Designing Path for Robot Arm Extensions Series with the Aim of Avoiding Obstruction with Recurring Neural Network

In this paper, recurrent neural network is used for path planning in the joint space of the robot with obstacle in the workspace of the robot. To design the neural network, first a performance index has been defined as sum of square of error tracking of final executor. Then, obstacle avoidance scheme is presented based on its space coordinate and its minimum distance between the obstacle and ea...

متن کامل

Application of Wavelet Neural Network in Forward Kinematics Solution of 6-RSU Co-axial Parallel Mechanism Based on Final Prediction Error

Application of artificial neural network (ANN) in forward kinematic solution (FKS) of a novel co-axial parallel mechanism with six degrees of freedom (6-DOF) is addressed in Current work. The mechanism is known as six revolute-spherical-universal (RSU) and constructed by 6-RSU co-axial kinematic chains in parallel form. First, applying geometrical analysis and vectorial principles the kinematic...

متن کامل

An efficient one-layer recurrent neural network for solving a class of nonsmooth optimization problems

Constrained optimization problems have a wide range of applications in science, economics, and engineering. In this paper, a neural network model is proposed to solve a class of nonsmooth constrained optimization problems with a nonsmooth convex objective function subject to nonlinear inequality and affine equality constraints. It is a one-layer non-penalty recurrent neural network based on the...

متن کامل

Kinematic Synthesis of Parallel Manipulator via Neural Network Approach

In this research, Artificial Neural Networks (ANNs) have been used as a powerful tool to solve the inverse kinematic equations of a parallel robot. For this purpose, we have developed the kinematic equations of a Tricept parallel kinematic mechanism with two rotational and one translational degrees of freedom (DoF). Using the analytical method, the inverse kinematic equations are solved for spe...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • CoRR

دوره abs/1705.01346  شماره 

صفحات  -

تاریخ انتشار 2017